Convergence results for the EM approach to mixtures of experts architectures
نویسندگان
چکیده
The Expectation-Maximization (EM) algorithm is an iterative approach to maximum likelihood parameter estimation. Jordan and Jacobs (1994) recently proposed an EM algorithm for the mixture of experts architecture of Jacobs, Jordan, Nowlan and Hinton (1991) and the hierarchical mixture of experts architecture of Jordan and Jacobs (1992). They showed empirically that the EM algorithm for these architectures yields signiicantly faster convergence than gradient ascent. In the current paper we provide a theoretical analysis of this algorithm. We show that the algorithm can be regarded as a variable metric algorithm with its searching direction having a positive projection on the gradient of the log likelihood. We also analyze the convergence of the algorithm and provide an explicit expression for the convergence rate. In addition, we describe an acceleration technique that yields a signiicant speedup in simulation experiments.
منابع مشابه
An Alternative Model for Mixtures of Experts
An alternative model is proposed for mixtures of experts by utiliz ing a di erent parametric form for the gating network The mod i ed model is trained by an EM algorithm In comparison with earlier models trained by either EM or gradient ascent there is no need to select a learning stepsize to guarantee the convergence of the learning procedure We report simulation experiments which show that th...
متن کاملAnalysis of Strategies for Integrating Humanities with Medical Sciences in Iran from the Perspective of Humanities Experts with Convergence Sciences Approach
Introduction: Medical humanities are one of the most important interdisciplinary needs. Since the most important discussion in this field is how to achieve it, the purpose of this study is to analyze it from the perspective of humanities experts in the Iran. Method: The present study is a national wide qualitative study that was conducted in 2019. Considering the saturation criterion, data wer...
متن کاملMixture of experts architectures for neural networks as a special case of conditional expectation formula
Recently a new interesting architecture of neural networks called “mixture of experts” has been proposed as a tool of real multivariate approximation or classification. It is shown that, in some cases, the underlying problem of prediction can be solved by estimating the joint probability density of involved variables. Assuming the model of Gaussian mixtures we can explictly write the optimal mi...
متن کاملSpectral Experts for Estimating Mixtures of Linear Regressions
Discriminative latent-variable models are typically learned using EM or gradient-based optimization, which suffer from local optima. In this paper, we develop a new computationally efficient and provably consistent estimator for a mixture of linear regressions, a simple instance of a discriminative latentvariable model. Our approach relies on a lowrank linear regression to recover a symmetric t...
متن کاملConvergence of the EM Algorithm for Gaussian Mixtures with Unbalanced Mixing Coefficients
The speed of convergence of the Expectation Maximization (EM) algorithm for Gaussian mixture model fitting is known to be dependent on the amount of overlap among the mixture components. In this paper, we study the impact of mixing coefficients on the convergence of EM. We show that when the mixture components exhibit some overlap, the convergence of EM becomes slower as the dynamic range among...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Neural Networks
دوره 8 شماره
صفحات -
تاریخ انتشار 1995